Optimal prior-free probabilistic variable selection in Gaussian linear regression

نویسندگان

  • Ryan Martin
  • Huiping Xu
  • Zuoyi Zhang
  • Chuanhai Liu
چکیده

Model selection is fundamental to scientific discovery, but a general framework that gives valid prior-free probabilistic assessments of the quality of individual models remains to be worked out. In this paper, we propose an inferential model (IM) approach to accomplish this for the special case of variable selection in a full-rank Gaussian linear regression model. This proposal is based on the notion of simultaneously valid inference for a collection of assertions of interest, and we develop some general results and characterizations of the corresponding optimal predictive random sets. An interesting and important by-product of our IM approach is that the resulting validity holds when the IM is projected down to a sub-model, even if that sub-model is selected by the data. In other words, the proposed IM approach automatically gives valid post-selection inference. An IM-driven variable selection procedure is proposed, and are shown to compare favorably to existing methods based on realand simulated-data examples.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Novel Radial Basis Function Neural Networks based on Probabilistic Evolutionary and Gaussian Mixture Model for Satellites Optimum Selection

In this study, two novel learning algorithms have been applied on Radial Basis Function Neural Network (RBFNN) to approximate the functions with high non-linear order. The Probabilistic Evolutionary (PE) and Gaussian Mixture Model (GMM) techniques are proposed to significantly minimize the error functions. The main idea is concerning the various strategies to optimize the procedure of Gradient ...

متن کامل

Support vector regression with random output variable and probabilistic constraints

Support Vector Regression (SVR) solves regression problems based on the concept of Support Vector Machine (SVM). In this paper, a new model of SVR with probabilistic constraints is proposed that any of output data and bias are considered the random variables with uniform probability functions. Using the new proposed method, the optimal hyperplane regression can be obtained by solving a quadrati...

متن کامل

Automatic Smoothing and Variable Selection via Regularization

This thesis focuses on developing computational methods and the general theory of automatic smoothing and variable selection via regularization. Methods of regularization are a commonly used technique to get stable solution to ill-posed problems such as nonparametric regression and classification. In recent years, methods of regularization have also been successfully introduced to address a cla...

متن کامل

Bayesian Density Regression through Projections

We explore a Bayesian density regression model driven by linear projections of covariates. This offers an alternative to variable selection and provides the best linear compression of covariates for predicting a response in a regression setting. We provide a detailed construction of our probabilistic model based on smooth Gaussian processes and uniformly distributed linear subspaces. We prove p...

متن کامل

Sparse oracle inequalities for variable selection via regularized quantization

We give oracle inequalities on procedures which combines quantization and variable selection via a weighted Lasso k-means type algorithm. The results are derived for a general family of weights, which can be tuned to size the influence of the variables in different ways. Moreover, these theoretical guarantees are proved to adapt the corresponding sparsity of the optimal codebooks, suggesting th...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2014